Goto

Collaborating Authors

 statistical and computational trade-off


[1.1. Motivations for regression with sparse interaction terms. ] Regression with interaction terms

Neural Information Processing Systems

We thank all the reviewers for the helpful comments. Here, we address the main concerns raised by the reviewers. While the traditional method (e.g., Lasso) can find important individual The motivations and application are discussed in [1.1.] When the sparse assumption doesn't hold] Theoretically, the sparsity assumption is commonly Motivations] The motivations and one real application where sparsity holds are discussed in [1.1.]


Statistical and Computational Trade-off in Multi-Agent Multi-Armed Bandits

Neural Information Processing Systems

We study the problem of regret minimization in Multi-Agent Multi-Armed Bandits (MAMABs) where the rewards are defined through a factor graph. We derive an instance-specific regret lower bound and characterize the minimal expected number of times each global action should be explored. Unfortunately, this bound and the corresponding optimal exploration process are obtained by solving a combinatorial optimization problem with a set of variables and constraints exponentially growing with the number of agents. We approximate the regret lower bound problem via Mean Field techniques to reduce the number of variables and constraints. By tuning the latter, we explore the trade-off between achievable regret and complexity. We devise Efficient Sampling for MAMAB (ESM), an algorithm whose regret asymptotically matches the corresponding approximated lower bound. We assess the regret and computational complexity of ESM numerically, using both synthetic and real-world experiments in radio communications networks.


Statistical and Computational Trade-Offs in Kernel K-Means

Neural Information Processing Systems

We investigate the efficiency of k-means in terms of both statistical and computational requirements. More precisely, we study a Nystr\om approach to kernel k-means. We analyze the statistical properties of the proposed method and show that it achieves the same accuracy of exact kernel k-means with only a fraction of computations. Indeed, we prove under basic assumptions that sampling $\sqrt{n}$ Nystr\om landmarks allows to greatly reduce computational costs without incurring in any loss of accuracy. To the best of our knowledge this is the first result showing in this kind for unsupervised learning.




Statistical and Computational Trade-off in Multi-Agent Multi-Armed Bandits

Neural Information Processing Systems

We study the problem of regret minimization in Multi-Agent Multi-Armed Bandits (MAMABs) where the rewards are defined through a factor graph. We derive an instance-specific regret lower bound and characterize the minimal expected number of times each global action should be explored. Unfortunately, this bound and the corresponding optimal exploration process are obtained by solving a combinatorial optimization problem with a set of variables and constraints exponentially growing with the number of agents. We approximate the regret lower bound problem via Mean Field techniques to reduce the number of variables and constraints. By tuning the latter, we explore the trade-off between achievable regret and complexity.


Reviews: Statistical and Computational Trade-Offs in Kernel K-Means

Neural Information Processing Systems

Summary The paper investigates the kernel k-means problem, proposing a new approximation of the method based on Nystrom embeddings. In particular, both the statistical accuracy and computational efficiency of the proposed approximation is studied. The main contribution is the derivation of a theoretical bound limiting the error cost; the proposed method can achieve such theoretical guarantee with a computational cost of (O(\sqrt{n})). In practice, this implies that using \sqrt{n} points (with n being the original sample size) is enough to obtain a good enough approximation. The authors also propose an approach to choose the points while ensuring a balance trade off between preserving a good approximation and a small complexity (dictionary size).


Statistical and Computational Trade-offs in Variational Inference: A Case Study in Inferential Model Selection

Bhatia, Kush, Kuang, Nikki Lijing, Ma, Yi-An, Wang, Yixin

arXiv.org Artificial Intelligence

Variational inference has recently emerged as a popular alternative to the classical Markov chain Monte Carlo (MCMC) in large-scale Bayesian inference. The core idea is to trade statistical accuracy for computational efficiency. In this work, we study these statistical and computational trade-offs in variational inference via a case study in inferential model selection. Focusing on Gaussian inferential models (or variational approximating families) with diagonal plus low-rank precision matrices, we initiate a theoretical study of the trade-offs in two aspects, Bayesian posterior inference error and frequentist uncertainty quantification error. From the Bayesian posterior inference perspective, we characterize the error of the variational posterior relative to the exact posterior. We prove that, given a fixed computation budget, a lower-rank inferential model produces variational posteriors with a higher statistical approximation error, but a lower computational error; it reduces variance in stochastic optimization and, in turn, accelerates convergence. From the frequentist uncertainty quantification perspective, we consider the precision matrix of the variational posterior as an uncertainty estimate, which involves an additional statistical error originating from the sampling uncertainty of the data. As a consequence, for small datasets, the inferential model need not be full-rank to achieve optimal estimation error (even with unlimited computation budget).


Statistical and Computational Trade-Offs in Kernel K-Means

Calandriello, Daniele, Rosasco, Lorenzo

Neural Information Processing Systems

We investigate the efficiency of k-means in terms of both statistical and computational requirements. More precisely, we study a Nystr\"om approach to kernel k-means. We analyze the statistical properties of the proposed method and show that it achieves the same accuracy of exact kernel k-means with only a fraction of computations. Indeed, we prove under basic assumptions that sampling $\sqrt{n}$ Nystr\"om landmarks allows to greatly reduce computational costs without incurring in any loss of accuracy. To the best of our knowledge this is the first result showing in this kind for unsupervised learning.


Statistical and Computational Trade-Offs in Kernel K-Means

Calandriello, Daniele, Rosasco, Lorenzo

arXiv.org Machine Learning

We investigate the efficiency of k-means in terms of both statistical and computational requirements. More precisely, we study a Nystr\"om approach to kernel k-means. We analyze the statistical properties of the proposed method and show that it achieves the same accuracy of exact kernel k-means with only a fraction of computations. Indeed, we prove under basic assumptions that sampling $\sqrt{n}$ Nystr\"om landmarks allows to greatly reduce computational costs without incurring in any loss of accuracy. To the best of our knowledge this is the first result of this kind for unsupervised learning.